171 research outputs found

    On the Relationship between Transmission Power and Capacity of an Underwater Acoustic Communication Channel

    Full text link
    The underwater acoustic channel is characterized by a path loss that depends not only on the transmission distance, but also on the signal frequency. As a consequence, transmission bandwidth depends on the transmission distance, a feature that distinguishes an underwater acoustic system from a terrestrial radio system. The exact relationship between power, transmission band, distance and capacity for the Gaussian noise scenario is a complicated one. This work provides a closed-form approximate model for 1) power consumption, 2) band-edge frequency and 3) bandwidth as functions of distance and capacity required for a data link. This approximate model is obtained by numerical evaluation of analytical results which takes into account physical models of acoustic propagation loss and ambient noise. The closed-form approximations may become useful tools in the design and analysis of underwater acoustic networks.Comment: 6 pages, 9 Figures, Awaiting acceptance to IEEE Oceans 08 (Conference), Kobe, Japa

    Broadcasting in Time-Division Duplexing: A Random Linear Network Coding Approach

    Full text link
    We study random linear network coding for broadcasting in time division duplexing channels. We assume a packet erasure channel with nodes that cannot transmit and receive information simultaneously. The sender transmits coded data packets back-to-back before stopping to wait for the receivers to acknowledge the number of degrees of freedom, if any, that are required to decode correctly the information. We study the mean time to complete the transmission of a block of packets to all receivers. We also present a bound on the number of stops to wait for acknowledgement in order to complete transmission with probability at least 1−ϵ1-\epsilon, for any ϵ>0\epsilon>0. We present analysis and numerical results showing that our scheme outperforms optimal scheduling policies for broadcast, in terms of the mean completion time. We provide a simple heuristic to compute the number of coded packets to be sent before stopping that achieves close to optimal performance with the advantage of a considerable reduction in the search time.Comment: 6 pages, 5 figures, Submitted to Workshop on Network Coding, Theory, and Applications (NetCod 2009

    Whether and Where to Code in the Wireless Relay Channel

    Full text link
    The throughput benefits of random linear network codes have been studied extensively for wirelined and wireless erasure networks. It is often assumed that all nodes within a network perform coding operations. In energy-constrained systems, however, coding subgraphs should be chosen to control the number of coding nodes while maintaining throughput. In this paper, we explore the strategic use of network coding in the wireless packet erasure relay channel according to both throughput and energy metrics. In the relay channel, a single source communicates to a single sink through the aid of a half-duplex relay. The fluid flow model is used to describe the case where both the source and the relay are coding, and Markov chain models are proposed to describe packet evolution if only the source or only the relay is coding. In addition to transmission energy, we take into account coding and reception energies. We show that coding at the relay alone while operating in a rateless fashion is neither throughput nor energy efficient. Given a set of system parameters, our analysis determines the optimal amount of time the relay should participate in the transmission, and where coding should be performed.Comment: 11 pages, 12 figures, to be published in the IEEE JSAC Special Issue on Theories and Methods for Advanced Wireless Relay

    Generalized Deduplication: Bounds, Convergence, and Asymptotic Properties

    Full text link
    We study a generalization of deduplication, which enables lossless deduplication of highly similar data and show that standard deduplication with fixed chunk length is a special case. We provide bounds on the expected length of coded sequences for generalized deduplication and show that the coding has asymptotic near-entropy cost under the proposed source model. More importantly, we show that generalized deduplication allows for multiple orders of magnitude faster convergence than standard deduplication. This means that generalized deduplication can provide compression benefits much earlier than standard deduplication, which is key in practical systems. Numerical examples demonstrate our results, showing that our lower bounds are achievable, and illustrating the potential gain of using the generalization over standard deduplication. In fact, we show that even for a simple case of generalized deduplication, the gain in convergence speed is linear with the size of the data chunks.Comment: 15 pages, 4 figures. This is the full version of a paper accepted for GLOBECOM 201

    Random Linear Network Coding For Time Division Duplexing: Energy Analysis

    Full text link
    We study the energy performance of random linear network coding for time division duplexing channels. We assume a packet erasure channel with nodes that cannot transmit and receive information simultaneously. The sender transmits coded data packets back-to-back before stopping to wait for the receiver to acknowledge the number of degrees of freedom, if any, that are required to decode correctly the information. Our analysis shows that, in terms of mean energy consumed, there is an optimal number of coded data packets to send before stopping to listen. This number depends on the energy needed to transmit each coded packet and the acknowledgment (ACK), probabilities of packet and ACK erasure, and the number of degrees of freedom that the receiver requires to decode the data. We show that its energy performance is superior to that of a full-duplex system. We also study the performance of our scheme when the number of coded packets is chosen to minimize the mean time to complete transmission as in [1]. Energy performance under this optimization criterion is found to be close to optimal, thus providing a good trade-off between energy and time required to complete transmissions.Comment: 5 pages, 6 figures, Accepted to ICC 200

    European Wireless 2019; 25th European Wireless Conference. Aarhus, Denmark

    Get PDF
    This paper describes a new design of Reed-Solomon (RS) codes when using composite extension fields. Our ultimate goal is to provide codes that remain Maximum Distance Separable (MDS), but that can be processed at higher speeds in the encoder and decoder. This is possible by using coefficients in the generator matrix that belong to smaller (and faster) finite fields of the composite extension and limiting the use of the larger (and slower) finite fields to a minimum. We provide formulae and an algorithm to generate such constructions starting from a Vandermonde RS generator matrix and show that even the simplest constructions, e.g., using only processing in two finite fields, can speed up processing by as much as two-fold compared to a Vandermonde RS and Cauchy RS while using the same decoding algorithm, and more than two-fold compared to other RS Cauchy and FFT-based RS

    Random Linear Network Coding For Time Division Duplexing: When To Stop Talking And Start Listening

    Full text link
    A new random linear network coding scheme for reliable communications for time division duplexing channels is proposed. The setup assumes a packet erasure channel and that nodes cannot transmit and receive information simultaneously. The sender transmits coded data packets back-to-back before stopping to wait for the receiver to acknowledge (ACK) the number of degrees of freedom, if any, that are required to decode correctly the information. We provide an analysis of this problem to show that there is an optimal number of coded data packets, in terms of mean completion time, to be sent before stopping to listen. This number depends on the latency, probabilities of packet erasure and ACK erasure, and the number of degrees of freedom that the receiver requires to decode the data. This scheme is optimal in terms of the mean time to complete the transmission of a fixed number of data packets. We show that its performance is very close to that of a full duplex system, while transmitting a different number of coded packets can cause large degradation in performance, especially if latency is high. Also, we study the throughput performance of our scheme and compare it to existing half-duplex Go-back-N and Selective Repeat ARQ schemes. Numerical results, obtained for different latencies, show that our scheme has similar performance to the Selective Repeat in most cases and considerable performance gain when latency and packet error probability is high.Comment: 9 pages, 9 figures, Submitted to INFOCOM'0

    Network Coding for Multi-Resolution Multicast

    Full text link
    Multi-resolution codes enable multicast at different rates to different receivers, a setup that is often desirable for graphics or video streaming. We propose a simple, distributed, two-stage message passing algorithm to generate network codes for single-source multicast of multi-resolution codes. The goal of this "pushback algorithm" is to maximize the total rate achieved by all receivers, while guaranteeing decodability of the base layer at each receiver. By conducting pushback and code generation stages, this algorithm takes advantage of inter-layer as well as intra-layer coding. Numerical simulations show that in terms of total rate achieved, the pushback algorithm outperforms routing and intra-layer coding schemes, even with codeword sizes as small as 10 bits. In addition, the performance gap widens as the number of receivers and the number of nodes in the network increases. We also observe that naiive inter-layer coding schemes may perform worse than intra-layer schemes under certain network conditions.Comment: 9 pages, 16 figures, submitted to IEEE INFOCOM 201
    • …
    corecore